Mitigating Negative Transfer in Multi-Task Learning with Exponential Moving Average Loss Weighting Strategies (Student Abstract)
نویسندگان
چکیده
Multi-Task Learning (MTL) is a growing subject of interest in deep learning, due to its ability train models more efficiently on multiple tasks compared using group conventional single-task models. However, MTL can be impractical as certain dominate training and hurt performance others, thus making some perform better model multi-task one. Such problems are broadly classified negative transfer, many prior approaches the literature have been made mitigate these issues. One such current approach alleviate transfer weight each losses so that they same scale. Whereas loss balancing rely either optimization or complex numerical analysis, none directly scale based their observed magnitudes. We propose techniques for scaling by exponential moving average benchmark them against best-performing methods three established datasets. On datasets, achieve comparable, if not higher, methods.
منابع مشابه
Move from Perturbed scheme to exponential weighting average
In an online decision problem, one makes decisions often with a pool of decisions’ sequence called experts but without knowledge of the future. After each step, one pays a cost based on the decision and observed rate. One reasonal goal would be to perform as well as the best expert in the pool. The modern and well-known way to attain this goal is the algorithm of exponential weighting. However,...
متن کاملMoving Average Processes with Infinite Variance
The sample autocorrelation function (acf) of a stationary process has played a central statistical role in traditional time series analysis, where the assumption is made that the marginal distribution has a second moment. Now, the classical methods based on acf are not applicable in heavy tailed modeling. Using the codifference function as dependence measure for such processes be shown it be as...
متن کاملExponential Moving Average Model in Parallel Speech Recognition Training
As training data rapid growth, large-scale parallel training with multi-GPUs cluster is widely applied in the neural network model learning currently. We present a new approach that applies exponential moving average method in large-scale parallel training of neural network model. It is a non-interference strategy that the exponential moving average model is not broadcasted to distributed worke...
متن کاملMulti-Task and Transfer Learning with Recurrent Neural Networks
Being able to obtain a model of a dynamical system is useful for a variety of purposes, e.g. condition monitoring and model-based control. The dynamics of complex technical systems such as gas or wind turbines can be approximated by data driven models, e.g. recurrent neural networks. Such methods have proven to be powerful alternatives to analytical models which are not always available or may ...
متن کاملAsymmetric Multi-task Learning Based on Task Relatedness and Loss
We propose a novel multi-task learning method that minimizes the effect of negative transfer by allowing asymmetric transfer between the tasks based on task relatedness as well as the amount of individual task losses, which we refer to as Asymmetric Multi-task Learning (AMTL). To tackle this problem, we couple multiple tasks via a sparse, directed regularization graph, that enforces each task p...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2023
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v37i13.26983